"main.py --dataset Computers --hidden_dim 2048 --epoch 10 --lr 0.0001 --runtime 1 --classifier_time 10 --classifier_epochs 6000 --cuda 1 --classifier_lr 0.1 --classifier_decay 0.00001 --batch_norm --loss cosmean --a1 0.7 --a2 0.3 --classifier_seed 0 --GCN_layers_1 1"
test_acc: 0.876453488372093
test_std: 0.011035987949260586
test_acc_mean: 0.876453488372093
test_acc_std: 0.0
{
  "classifier_epochs": 6000,
  "batch_size": 1,
  "patience": 200,
  "l2_coef": 0.0,
  "drop_prob": 0.0,
  "dataset": "Computers",
  "num_hop": 0,
  "cuda": 1,
  "seed": 1234,
  "classifier_lr": 0.1,
  "classifier_decay": 1e-05,
  "classifier_seed": 0,
  "classifier_time": 10,
  "classifier": null,
  "hidden_dim": 2048,
  "proj_dim": 256,
  "fc_out": false,
  "input_dropout": 0.4,
  "dropout": 0.4,
  "batch_norm": true,
  "optimizer": "adam",
  "lr": 0.0001,
  "decay": 0.0005,
  "epoch": 10,
  "alpha_dim": "sc",
  "no_alpha_sigmoid": false,
  "beta_dim": "sc",
  "block": "constant",
  "function": "laplacian",
  "use_mlp": false,
  "add_source": false,
  "time": 3.0,
  "augment": false,
  "method": "euler",
  "step_size": 1.0,
  "max_iters": 100,
  "adjoint_method": "adaptive_heun",
  "adjoint": false,
  "adjoint_step_size": 1,
  "tol_scale": 1.0,
  "tol_scale_adjoint": 1.0,
  "ode_blocks": 1,
  "max_nfe": 1000,
  "no_early": false,
  "earlystopxT": 3,
  "max_test_steps": 100,
  "beltrami": false,
  "leaky_relu_slope": 0.2,
  "attention_dropout": 0.0,
  "heads": 4,
  "attention_norm_idx": 0,
  "attention_dim": 16,
  "mix_features": false,
  "reweight_attention": false,
  "attention_type": "scaled_dot",
  "square_plus": false,
  "data_norm": "gcn",
  "self_loop_weight": 1.0,
  "runtime": 1,
  "loss": "cosmean",
  "random_splits": false,
  "alpha_ode": 1.0,
  "alpha_ode_1": 1.0,
  "alpha_ode_2": 0.1,
  "a1": 0.7,
  "a2": 0.3,
  "reaction_term": "bspm",
  "beta_diag": false,
  "lambda": 0.1,
  "GCN_layers_1": 1,
  "GCN_layers_2": 2,
  "feature_shuffled": false,
  "num_nodes": 13752
}
